nome: Rafael Marinho Ferreira

email: rafamarinho87@gmail.com

github: https://github.com/rferreira13

In [1]:
import pandas as pd
import numpy as np
import scipy.stats as st
import seaborn as sns
import matplotlib.pyplot as plt

import statsmodels.api as sm
import statsmodels.formula.api as smf
from statsmodels.tsa.stattools import adfuller

from sklearn.ensemble import RandomForestClassifier, ExtraTreesClassifier, GradientBoostingClassifier, HistGradientBoostingClassifier
from sklearn.svm import SVC, LinearSVC
from sklearn.linear_model import LogisticRegression
from sklearn.neighbors import KNeighborsClassifier
from imblearn.over_sampling import RandomOverSampler
from sklearn.model_selection import GridSearchCV
from sklearn.preprocessing import StandardScaler
from sklearn.metrics import classification_report, confusion_matrix, f1_score
from sklearn.model_selection import train_test_split
from sklearn.metrics import make_scorer, recall_score
In [2]:
df = pd.read_excel("../Test O_G_Equipment_Data.xlsx")

Major task¶

  • Investigate one piece of equipment in different time cycles to understand what characteristics and parameters of the sensors might indicate that the equipment is on the verge of failing.

Task 1¶

  • Calculate how many times the equipment has failed.

During the FPSO’s operation, various factors can cause the machine to fail and prolong its failure state.

We ask you to explore the available data, identify, and calculate the number of times the equipment has failed throughout its operation.

In [3]:
counts = df["Fail"].value_counts(dropna=False).rename_axis("Fail").reset_index(name="count")
fig, ax = plt.subplots(figsize=(8,6))
bars = ax.bar(counts["Fail"].astype(str), counts["count"])
ax.bar_label(bars, padding=3)
ax.set_title("Count of Fail and Not Fail Events", fontsize=26, pad=20, loc="center")
ax.set_ylim(0, df.shape[0])
fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
plt.show()
No description has been provided for this image
In [4]:
"{:.2%}".format(df['Fail'].mean())
Out[4]:
'8.25%'

Task 2¶

  • Categorize equipment failures by setup configurations (Preset 1 and Preset 2).

  • How do the variables Preset_1 and Preset_2 behave during operation?

  • What insights can we derive from these variables?

In [5]:
def failure_report(data, group):
    data_copy = data.copy()
    total = (data_copy[group].astype(str)
            .value_counts().rename_axis(group)
            .reset_index(name='Number of observations'))
    taxa = (
        data_copy[data_copy["Fail"]][group].astype(str).value_counts() 
        / data_copy[group].astype(str).value_counts()
        ).reset_index(name='Failure percentage')
    base = (total.merge(taxa, on=group, how='left')
                .fillna({'Failure percentage': 0})
                .sort_values(group))
    fig, ax1 = plt.subplots(figsize=(8,6))
    bars = ax1.bar(base[group].astype(str), base['Number of observations'], label='Number of observations')
    ax1.bar_label(bars, padding=3)
    ax1.set_ylabel('Number of observations')
    ax1.set_xlabel(group)
    ax2 = ax1.twinx()
    ax2.plot(base[group].astype(str), base['Failure percentage'], marker='o', linewidth=2, color='red', label='Failure percentage')
    for i, v in enumerate(base['Failure percentage']):
        ax2.text(i, v, f'{v:.1%}', ha='center', va='bottom')
    ax2.set_ylabel('Failure (%)')
    ax1.set_title(f"Number of observations versus Failure percentage\nConsidering {group}", fontsize=26, pad=20)
    ticks = ax2.get_yticks()
    ax2.set_yticklabels([f'{t:.1%}' for t in ticks])
    fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
    plt.show()

def heatmap_failure(data):
    data_copy = data.copy()
    matrix_data = (data_copy.groupby(['Preset_1','Preset_2'])['Fail'].mean()
        .mul(100).unstack(fill_value=0).sort_index())
    matrix_data.columns = matrix_data.columns.astype(str)
    matrix_data.index = matrix_data.index.astype(str)
    fig, ax = plt.subplots(figsize=(8,6))
    im = ax.imshow(matrix_data.values, cmap='Reds', aspect='auto')
    ax.set_xticks(range(matrix_data.shape[1]))
    ax.set_xticklabels(matrix_data.columns)
    ax.set_yticks(range(matrix_data.shape[0]))
    ax.set_yticklabels(matrix_data.index)
    for i in range(matrix_data.shape[0]):
        for j in range(matrix_data.shape[1]):
            ax.text(j, i, f"{matrix_data.values[i, j]:.1f}", ha='center', va='center')
    ax.set_xlabel('Preset_2')
    ax.set_ylabel('Preset_1')
    cbar = plt.colorbar(im)
    cbar.set_label('Fail %')
    ax.set_title("Percentage Failure Heatmap\nCombining Preset_1 and Preset_2", fontsize=26, pad=20)
    fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
    plt.show()

def heatmap_configuration_changes(data, group):
    data_copy = data.copy()
    data_copy[f'Previous {group}'] = data_copy[group].shift()
    data_copy = data_copy.dropna()
    data_copy[f'Previous {group}'] = data_copy[f'Previous {group}'].astype(int)
    data_copy[f'Previous {group}'] = data_copy[f'Previous {group}'].astype(str)
    data_copy[group] = data_copy[group].astype(str)
    n_changes_df = (
        data_copy
        [[f'Previous {group}', group]]
        .value_counts()
        .reset_index(name="Number of Changes")
    )
    n_changes_df["Percent of changes during operation"] = n_changes_df["Number of Changes"] / data_copy.shape[0]
    n_changes_df = (
        n_changes_df
        .rename(columns={
            f'Previous {group}':"From",
            group:"To"
            }
        )
        .sort_values(by=["From", "To"])
        .reset_index(drop=True)
    )
    order_x = sorted(n_changes_df["From"].unique())
    order_y = sorted(n_changes_df["To"].unique())
    fig, ax = plt.subplots(figsize=(8,6))
    z = np.zeros((len(order_y), len(order_x)))
    for i, y in enumerate(order_y):
        for j, x in enumerate(order_x):
            v = n_changes_df.loc[(n_changes_df["From"]==x) & (n_changes_df["To"]==y), "Percent of changes during operation"].values
            z[i, j] = v[0] if len(v) else 0
    vmax = z.max() if z.max() > 0 else 1
    vmin = z[z>0].min() if np.any(z>0) else 0
    im = ax.imshow(z, cmap="Reds", aspect="auto", vmin=vmin, vmax=vmax)
    ax.set_xticks(range(len(order_x)))
    ax.set_xticklabels(order_x)
    ax.set_yticks(range(len(order_y)))
    ax.set_yticklabels(order_y)
    for i in range(len(order_y)):
        for j in range(len(order_x)):
            ax.text(j, i, f"{z[i,j]:.2%}", ha="center", va="center")
    ticks = np.linspace(vmin, vmax, 6)
    cbar = plt.colorbar(im, ticks=ticks)
    cbar.set_ticklabels([f"{t:.0%}" for t in ticks])
    cbar.set_label("Percent of changes during operation")
    ax.set_title(f"Percent of changes from {group} configurations during operation", fontsize=26, pad=20)
    fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
    if n_changes_df.shape[0] > 200:
        fig.set_figheight(9)
        fig.set_figwidth(20)
    plt.show()




def boxplots(data, group, variable):
    data_copy = data.copy()
    model = smf.ols(f'{variable} ~ C({group})', data=data_copy).fit()
    resid = model.resid
    fitted = model.fittedvalues
    fig = st.probplot(resid, dist="norm", plot=plt)
    plt.title(f"QQ plot of residuals - {variable} ~ C({group})")
    plt.show()
    plt.scatter(fitted, resid, s=8, alpha=0.6)
    plt.axhline(0, lw=1, color='k')
    plt.xlabel("Fitted values")
    plt.ylabel("Residuals")
    plt.title(f"Residuals vs Fitted - {variable} ~ C({group})")
    plt.show()
    fig, ax = plt.subplots()
    data_copy.boxplot(column=variable, by=group, ax=ax)
    ax.set_xlabel(group)
    ax.set_ylabel(variable)
    plt.suptitle("")
    plt.xticks(rotation=45)
    fig.subplots_adjust(top=0.85, bottom=0.2)
    model = smf.ols(f'{variable} ~ C({group})', data=data_copy).fit()
    anova_tbl = sm.stats.anova_lm(model, typ=2)
    df1 = float(anova_tbl.loc[f'C({group})', 'df'])
    df2 = float(anova_tbl.loc['Residual', 'df'])
    Fval = float(anova_tbl.loc[f'C({group})', 'F'])
    pval = float(anova_tbl.loc[f'C({group})', 'PR(>F)'])
    ss_effect = float(anova_tbl.loc[f'C({group})', 'sum_sq'])
    ss_error  = float(anova_tbl.loc['Residual', 'sum_sq'])
    mse = ss_error / df2
    y = data_copy[variable].to_numpy(dtype=float)
    ss_total = float(((y - y.mean())**2).sum())
    eta2   = ss_effect / ss_total
    omega2 = (ss_effect - df1 * mse) / (ss_total + mse)
    k = data_copy[group].nunique()
    N = len(data_copy)
    text_ = (
        f"ANOVA (one-way) — {variable} ~ {group}\n"
        f"F({int(df1)}, {int(df2)}) = {Fval:.2f}, p = {pval:.3g} "
        f"| η² = {eta2:.3f}, ω² = {omega2:.3f} "
        f"| k = {k}, N = {N}"
    )
    fig.text(0.0, 0.98, text_, ha='left', va='top', bbox=dict(facecolor='white', alpha=0.75, edgecolor='black', linewidth=1))
    plt.show()

Evaluating Presets regarding Number of obsertations and Failure percentage¶

In [6]:
failure_report(df, "Preset_1")
failure_report(df, "Preset_2")

aux_data = df.copy()
aux_data["Preset_1_2"] = df['Preset_1'].astype(str) + df['Preset_2'].astype(str)

failure_report(aux_data, "Preset_1_2")
/tmp/ipykernel_28877/1418244333.py:25: UserWarning: set_ticklabels() should only be used with a fixed number of ticks, i.e. after set_ticks() or using a FixedLocator.
  ax2.set_yticklabels([f'{t:.1%}' for t in ticks])
No description has been provided for this image
/tmp/ipykernel_28877/1418244333.py:25: UserWarning: set_ticklabels() should only be used with a fixed number of ticks, i.e. after set_ticks() or using a FixedLocator.
  ax2.set_yticklabels([f'{t:.1%}' for t in ticks])
No description has been provided for this image
/tmp/ipykernel_28877/1418244333.py:25: UserWarning: set_ticklabels() should only be used with a fixed number of ticks, i.e. after set_ticks() or using a FixedLocator.
  ax2.set_yticklabels([f'{t:.1%}' for t in ticks])
No description has been provided for this image

Presets Failure Percentage Heatmap¶

In [7]:
heatmap_failure(df)
No description has been provided for this image

Presets state changes during operation¶

In [8]:
heatmap_configuration_changes(df, "Preset_1")
heatmap_configuration_changes(df, "Preset_2")

aux_data = df.copy()
aux_data["Preset_1_2"] = df['Preset_1'].astype(str) + df['Preset_2'].astype(str)

heatmap_configuration_changes(aux_data, "Preset_1_2")
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image

Boxplots per presets during operation¶

In [9]:
for column in df.columns[3:-1]:
    boxplots(df, 'Preset_1', column)

for column in df.columns[3:-1]:
    boxplots(df, 'Preset_2', column)

aux = df.copy()
aux['Preset_1_2'] = aux['Preset_1'].astype(str) + '-' + aux['Preset_2'].astype(str)
for column in df.columns[3:-1]:
    boxplots(aux, 'Preset_1_2', column)
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image

Task 3¶

  • Categorize equipment failures by their nature/root cause according to parameter readings (temperature, pressure, and others).

  • Analyze patterns in these readings that could indicate specific failure types.

  • How do these patterns differ across operational regimes?

  • Provide insights based on your findings.

In [10]:
def cycles_per_variable(data, x_axis, variable):
    data_copy = data.copy()
    data_copy[variable].quantile(0.95)
    data_copy[variable].quantile(0.05)
    p05 = data_copy[variable].quantile(0.05)
    p95 = data_copy[variable].quantile(0.95)
    median = data_copy[variable].median()
    df_sorted = data_copy.sort_values([x_axis])
    fig, ax = plt.subplots(figsize=(8,6))
    ax.plot(df_sorted[x_axis], df_sorted[variable], color="red", linewidth=2, label=variable)
    ax.fill_between(df_sorted[x_axis], p05, p95, alpha=0.2, label='P05–P95')
    ax.axhline(median, linestyle='--', linewidth=2, color='blue', label='Median')
    ax.axhline(p95, linestyle=':', linewidth=1)
    ax.axhline(p05, linestyle=':', linewidth=1)
    for x in data_copy.loc[data_copy["Fail"], x_axis].unique():
        ax.axvline(x=x, linewidth=2, color="black", alpha=0.5)
    ax.plot([], [], color="black", alpha=0.5, linewidth=1, label="Failure events")
    ax.set_ylim(0, 1.1 * data_copy[variable].max())
    ax.set_title(f"{variable} per {x_axis} with Failure events", fontsize=26, pad=20)
    ax.set_xlabel(x_axis)
    ax.set_ylabel(variable)
    ax.legend()
    fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
    plt.show()

def failure_3d_plot(data, x_value, y_value, z_value):
    data_copy = data.copy()
    fig = plt.figure(figsize=(8,6))
    ax = fig.add_subplot(111, projection='3d')
    d_true = data_copy[data_copy["Fail"]==True]
    d_false = data_copy[data_copy["Fail"]==False]
    ax.scatter(d_true[x_value], d_true[y_value], d_true[z_value], c="black", s=3, label="True")
    ax.scatter(d_false[x_value], d_false[y_value], d_false[z_value], c="lightgray", s=3, label="False")
    ax.set_title(f"3d Plot {x_value} versus {y_value} versus {z_value} per Failure event", fontsize=26, pad=20)
    ax.set_xlabel(x_value)
    ax.set_ylabel(y_value)
    ax.set_zlabel(z_value)
    ax.legend(title="Fail")
    fig.subplots_adjust(left=0.0, right=1.0, top=0.9, bottom=0.0)
    plt.show()

def failure_2d_plot(data, x_value, y_value):
    data_copy = data.copy()
    fig, ax = plt.subplots(figsize=(8,6))
    d_true = data_copy[data_copy["Fail"]==True]
    d_false = data_copy[data_copy["Fail"]==False]
    ax.scatter(d_true[x_value], d_true[y_value], c="black", s=20, label="True")
    ax.scatter(d_false[x_value], d_false[y_value], c="lightgray", s=20, label="False")
    ax.set_title(f"Scatter Plot {x_value} versus {y_value} per Failure event", fontsize=26, pad=20)
    ax.set_xlabel(x_value)
    ax.set_ylabel(y_value)
    ax.legend(title="Fail")
    fig.subplots_adjust(left=0.15, right=0.95, top=0.85, bottom=0.15)
    plt.show()
In [11]:
for column in df.columns[3:-1]:
    cycles_per_variable(df, "Cycle", column)

aux_data = df.copy()

aux_data["Module Vibration"] = aux_data.apply(lambda x: ((x["VibrationX"]**2) + (x["VibrationY"]**2) + (x["VibrationZ"]**2))**(1/2), axis=1)

cycles_per_variable(aux_data, "Cycle", "Module Vibration")

failure_2d_plot(df, "Temperature", "Pressure")

failure_2d_plot(aux_data, "Module Vibration", "Frequency")

failure_2d_plot(aux_data, "Module Vibration", "Temperature")

failure_2d_plot(aux_data, "Module Vibration", "Pressure")

failure_3d_plot(df, "Temperature", "Pressure", "Frequency")

failure_3d_plot(df, "VibrationX", "VibrationY", "VibrationZ")
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
No description has been provided for this image
In [12]:
def test_stationarity(series, scope):

    print(f'Dickey-Fuller (Dataset {scope}):')
    adf_test = adfuller(series.dropna())

    adf_stat = adf_test[0]
    p_value = adf_test[1]
    critical_values = adf_test[4]


    print(f'ADF statistic: {adf_stat}')
    print(f'p-value: {p_value}')
    print('Critical values:')
    for key, value in critical_values.items():
        print(f'\t{key}: {value}')


    if adf_stat < critical_values['5%']:
        print("The series is stationary (Reject the null hypothesis)")
    else:
        print("The series is NOT stationary (Fail to reject the null hypothesis)")


numerical_features = ["Temperature", "Pressure", "VibrationX", "VibrationY", "VibrationZ", "Frequency"]
for col in numerical_features:
    s = df.sort_values('Cycle')[col]
    test_stationarity(s, col)
Dickey-Fuller (Dataset Temperature):
ADF statistic: -12.988680950011224
p-value: 2.8425779537946296e-24
Critical values:
	1%: -3.438581476199162
	5%: -2.865173218890781
	10%: -2.56870466056054
The series is stationary (Reject the null hypothesis)
Dickey-Fuller (Dataset Pressure):
ADF statistic: -8.973295804703648
p-value: 7.6502921761507e-15
Critical values:
	1%: -3.438602251755426
	5%: -2.8651823762743245
	10%: -2.5687095387840673
The series is stationary (Reject the null hypothesis)
Dickey-Fuller (Dataset VibrationX):
ADF statistic: -6.9958169063199565
p-value: 7.538878106337241e-10
Critical values:
	1%: -3.438623132449471
	5%: -2.8651915799370014
	10%: -2.568714441670417
The series is stationary (Reject the null hypothesis)
Dickey-Fuller (Dataset VibrationY):
ADF statistic: -8.009749176554534
p-value: 2.2177469565403674e-12
Critical values:
	1%: -3.438602251755426
	5%: -2.8651823762743245
	10%: -2.5687095387840673
The series is stationary (Reject the null hypothesis)
Dickey-Fuller (Dataset VibrationZ):
ADF statistic: -6.648510831277863
p-value: 5.191805426666831e-09
Critical values:
	1%: -3.438633612472885
	5%: -2.865196199232788
	10%: -2.5687169024206713
The series is stationary (Reject the null hypothesis)
Dickey-Fuller (Dataset Frequency):
ADF statistic: -9.59674762354758
p-value: 1.9690415765210325e-16
Critical values:
	1%: -3.438602251755426
	5%: -2.8651823762743245
	10%: -2.5687095387840673
The series is stationary (Reject the null hypothesis)
In [13]:
def kde(df, numerical_columns, target_col):

    num_cols = len(numerical_columns)
    plt.figure(figsize=(18, 15))

    for i, column in enumerate(numerical_columns):
        plt.subplot(num_cols, 2, 2*i+1)
        sns.kdeplot(data=df, x=column, hue=target_col, fill=True)
        plt.title(f'KDE distribution of {column} by {target_col}')


    plt.tight_layout()
    plt.show()

numerical_features = ["Temperature", "Pressure", "VibrationX", "VibrationY", "VibrationZ", "Frequency"]

kde(df, numerical_features, 'Fail')
No description has been provided for this image
In [14]:
def plot_correlation_heatmap(df,features):
    corr_matrix = df[features].corr()

    plt.figure(figsize=(8, 3))

    sns.heatmap(corr_matrix, annot=True, cmap='coolwarm', fmt='.2f', linewidths=0.5)

    plt.title('Heatmap of numerical features correlation')
    plt.show()

plot_correlation_heatmap(df,numerical_features)
No description has been provided for this image
In [15]:
def preprocess(data,drops, columns, target):
  data_copy = data.copy()
  y = data_copy[target]
  X = data_copy.drop(drops, axis=1)

  scaler = StandardScaler()
  X[columns] = scaler.fit_transform(X[columns])

  X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=42)
  return  X_train, X_test, y_train, y_test

def train_model(X_train, y_train, model_exp):

    scorer = make_scorer(recall_score, pos_label=1)

    if model_exp == 'Random Forest':
        model = RandomForestClassifier(random_state=42)
        param_grid = {
            'n_estimators': [100, 200, 300],
            'max_depth': [10, 20, None],
            'min_samples_split': [2, 5, 10]
        }

    elif model_exp == 'SVM':
        model = SVC(kernel='rbf', random_state=42)
        param_grid = {
            'C': [0.1, 1, 10],
            'gamma': ['scale', 'auto'],
            'kernel': ['rbf']
        }

    elif model_exp == 'Logistic Regression':
        model = LogisticRegression(max_iter=1000, solver='saga', n_jobs=-1)
        param_grid = {
            'C': [0.01, 0.1, 1, 10],
            'penalty': ['l1', 'l2']
        }

    elif model_exp == 'Extra Trees':
        model = ExtraTreesClassifier(random_state=42, n_jobs=-1)
        param_grid = {
            'n_estimators': [200, 400],
            'max_depth': [None, 10, 20],
            'min_samples_split': [2, 10]
        }

    elif model_exp == 'Gradient Boosting':
        model = GradientBoostingClassifier(random_state=42)
        param_grid = {
            'n_estimators': [100, 300],
            'learning_rate': [0.05, 0.1],
            'max_depth': [2, 3]
        }

    elif model_exp == 'HistGB':
        model = HistGradientBoostingClassifier(random_state=42)
        param_grid = {
            'max_depth': [None, 6, 12],
            'learning_rate': [0.05, 0.1],
            'max_leaf_nodes': [31, 63]
        }

    elif model_exp == 'KNN':
        model = KNeighborsClassifier()
        param_grid = {
            'n_neighbors': [5, 15, 30],
            'weights': ['uniform', 'distance'],
            'p': [1, 2]
        }

    elif model_exp == 'Linear SVM':
        model = LinearSVC(random_state=42)
        param_grid = {
            'C': [0.1, 1, 10]
        }

    grid_search = GridSearchCV(estimator=model, param_grid=param_grid,
                               scoring=scorer, cv=5, n_jobs=-1, verbose=2)
    grid_search.fit(X_train, y_train)

    print(f"Better parameters for {model_exp}: {grid_search.best_params_}")
    return grid_search.best_estimator_


def predict_and_evaluate(model, X_test, y_test):

  y_pred = model.predict(X_test)
  conf_matrix = confusion_matrix(y_test, y_pred)
  sns.heatmap(conf_matrix, annot=True, fmt='d')
  plt.title('Confusion Matrix')
  plt.show()
  print(f'Model: {model}')
  print(classification_report(y_test, y_pred))

  f1 = f1_score(y_test, y_pred)
  print(f'F1 score: {f1 * 100:.2f}%')
In [16]:
df["Preset_1"] = df["Preset_1"].astype(str)
df["Preset_2"] = df["Preset_2"].astype(str)
df_hot = pd.get_dummies(df, columns=["Preset_1", "Preset_2"], drop_first=False)

Task 4¶

  • Create a model (or models) using the technique you think is most appropriate and measure its performance.

  • Based on the given time-series dataset, which models or techniques are suitable for predicting whether the equipment will fail before it occurs?

  • Additionally, how can the model's performance be tuned and measured for this task?

In [17]:
models = ['Random Forest', 'SVM', 'Logistic Regression', 'Extra Trees', 'Gradient Boosting', 'HistGB', 'KNN', 'Linear SVM']
aux = df.copy()
aux["Fail"] = aux["Fail"].apply(lambda x: 1 if x else 0)
drops = ["Preset_1",  "Preset_2", "Fail", "Cycle"]
X_train, X_test, y_train, y_test = preprocess(aux, drops, numerical_features, "Fail")
ros = RandomOverSampler(sampling_strategy=0.5, random_state=42)
X_train_ros, y_train_ros = ros.fit_resample(X_train, y_train)
X_train_hot, X_test_hot, y_train_hot, y_test_hot = preprocess(df_hot, ["Fail", "Cycle"], numerical_features, "Fail")
X_train_hot_ros, y_train_hot_ros = ros.fit_resample(X_train_hot, y_train_hot)

for model in models:
  print('-------------------------------------------------------------------------------')
  print('-------------------------- Original dataset -----------------------------------')
  print(f'Model: {model}')
  print('-------------------------------------------------------------------------------')
  tmodel = train_model(X_train, y_train, model)
  predict_and_evaluate(tmodel, X_test, y_test)
  print('-------------------------------------------------------------------------------')
  print('-------------------------- Random oversampled ---------------------------------')
  print(f'Model: {model}')
  print('-------------------------------------------------------------------------------')
  tmodel = train_model(X_train_ros, y_train_ros, model)
  predict_and_evaluate(tmodel, X_test, y_test)
  print('-------------------------------------------------------------------------------')
  print('-------------------------- Original dataset one hot encoded -------------------')
  print(f'Model: {model}')
  print('-------------------------------------------------------------------------------')
  tmodel = train_model(X_train_hot, y_train_hot, model)
  predict_and_evaluate(tmodel, X_test_hot, y_test_hot)
  print('-------------------------------------------------------------------------------')
  print('-------------------------- Random oversampled one hot encoded -----------------')
  print(f'Model: {model}')
  print('-------------------------------------------------------------------------------')
  tmodel = train_model(X_train_hot_ros, y_train_hot_ros, model)
  predict_and_evaluate(tmodel, X_test_hot, y_test_hot)
  print('-------------------------------------------------------------------------------')
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: Random Forest
-------------------------------------------------------------------------------
Fitting 5 folds for each of 27 candidates, totalling 135 fits
Better parameters for Random Forest: {'max_depth': 10, 'min_samples_split': 2, 'n_estimators': 200}
No description has been provided for this image
Model: RandomForestClassifier(max_depth=10, n_estimators=200, random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.99      0.98       225
           1       0.70      0.47      0.56        15

    accuracy                           0.95       240
   macro avg       0.83      0.73      0.77       240
weighted avg       0.95      0.95      0.95       240

F1 score: 56.00%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: Random Forest
-------------------------------------------------------------------------------
Fitting 5 folds for each of 27 candidates, totalling 135 fits
Better parameters for Random Forest: {'max_depth': 10, 'min_samples_split': 2, 'n_estimators': 100}
No description has been provided for this image
Model: RandomForestClassifier(max_depth=10, random_state=42)
              precision    recall  f1-score   support

           0       0.98      0.97      0.97       225
           1       0.59      0.67      0.62        15

    accuracy                           0.95       240
   macro avg       0.78      0.82      0.80       240
weighted avg       0.95      0.95      0.95       240

F1 score: 62.50%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: Random Forest
-------------------------------------------------------------------------------
Fitting 5 folds for each of 27 candidates, totalling 135 fits
Better parameters for Random Forest: {'max_depth': 10, 'min_samples_split': 2, 'n_estimators': 100}
No description has been provided for this image
Model: RandomForestClassifier(max_depth=10, random_state=42)
              precision    recall  f1-score   support

       False       0.97      0.99      0.98       225
        True       0.78      0.47      0.58        15

    accuracy                           0.96       240
   macro avg       0.87      0.73      0.78       240
weighted avg       0.95      0.96      0.95       240

F1 score: 58.33%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: Random Forest
-------------------------------------------------------------------------------
Fitting 5 folds for each of 27 candidates, totalling 135 fits
Better parameters for Random Forest: {'max_depth': 10, 'min_samples_split': 2, 'n_estimators': 100}
No description has been provided for this image
Model: RandomForestClassifier(max_depth=10, random_state=42)
              precision    recall  f1-score   support

       False       0.97      0.98      0.97       225
        True       0.62      0.53      0.57        15

    accuracy                           0.95       240
   macro avg       0.79      0.76      0.77       240
weighted avg       0.95      0.95      0.95       240

F1 score: 57.14%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 6 candidates, totalling 30 fits
Better parameters for SVM: {'C': 10, 'gamma': 'scale', 'kernel': 'rbf'}
No description has been provided for this image
Model: SVC(C=10, random_state=42)
              precision    recall  f1-score   support

           0       0.99      0.97      0.98       225
           1       0.63      0.80      0.71        15

    accuracy                           0.96       240
   macro avg       0.81      0.88      0.84       240
weighted avg       0.96      0.96      0.96       240

F1 score: 70.59%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 6 candidates, totalling 30 fits
Better parameters for SVM: {'C': 10, 'gamma': 'scale', 'kernel': 'rbf'}
No description has been provided for this image
Model: SVC(C=10, random_state=42)
              precision    recall  f1-score   support

           0       0.99      0.95      0.97       225
           1       0.54      0.87      0.67        15

    accuracy                           0.95       240
   macro avg       0.77      0.91      0.82       240
weighted avg       0.96      0.95      0.95       240

F1 score: 66.67%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 6 candidates, totalling 30 fits
Better parameters for SVM: {'C': 10, 'gamma': 'scale', 'kernel': 'rbf'}
No description has been provided for this image
Model: SVC(C=10, random_state=42)
              precision    recall  f1-score   support

       False       0.96      0.97      0.97       225
        True       0.54      0.47      0.50        15

    accuracy                           0.94       240
   macro avg       0.75      0.72      0.73       240
weighted avg       0.94      0.94      0.94       240

F1 score: 50.00%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 6 candidates, totalling 30 fits
Better parameters for SVM: {'C': 10, 'gamma': 'scale', 'kernel': 'rbf'}
No description has been provided for this image
Model: SVC(C=10, random_state=42)
              precision    recall  f1-score   support

       False       0.96      0.97      0.97       225
        True       0.54      0.47      0.50        15

    accuracy                           0.94       240
   macro avg       0.75      0.72      0.73       240
weighted avg       0.94      0.94      0.94       240

F1 score: 50.00%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: Logistic Regression
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Logistic Regression: {'C': 1, 'penalty': 'l1'}
No description has been provided for this image
Model: LogisticRegression(C=1, max_iter=1000, n_jobs=-1, penalty='l1', solver='saga')
              precision    recall  f1-score   support

           0       0.97      0.97      0.97       225
           1       0.60      0.60      0.60        15

    accuracy                           0.95       240
   macro avg       0.79      0.79      0.79       240
weighted avg       0.95      0.95      0.95       240

F1 score: 60.00%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: Logistic Regression
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Logistic Regression: {'C': 0.1, 'penalty': 'l2'}
No description has been provided for this image
Model: LogisticRegression(C=0.1, max_iter=1000, n_jobs=-1, solver='saga')
              precision    recall  f1-score   support

           0       1.00      0.94      0.97       225
           1       0.52      1.00      0.68        15

    accuracy                           0.94       240
   macro avg       0.76      0.97      0.82       240
weighted avg       0.97      0.94      0.95       240

F1 score: 68.18%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: Logistic Regression
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
Better parameters for Logistic Regression: {'C': 10, 'penalty': 'l1'}
No description has been provided for this image
Model: LogisticRegression(C=10, max_iter=1000, n_jobs=-1, penalty='l1', solver='saga')
              precision    recall  f1-score   support

       False       0.97      0.97      0.97       225
        True       0.56      0.60      0.58        15

    accuracy                           0.95       240
   macro avg       0.77      0.78      0.78       240
weighted avg       0.95      0.95      0.95       240

F1 score: 58.06%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: Logistic Regression
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Logistic Regression: {'C': 0.1, 'penalty': 'l2'}
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
/opt/ds_test_docs_env/lib/python3.10/site-packages/sklearn/linear_model/_sag.py:348: ConvergenceWarning: The max_iter was reached which means the coef_ did not converge
  warnings.warn(
No description has been provided for this image
Model: LogisticRegression(C=0.1, max_iter=1000, n_jobs=-1, solver='saga')
              precision    recall  f1-score   support

       False       1.00      0.94      0.97       225
        True       0.54      1.00      0.70        15

    accuracy                           0.95       240
   macro avg       0.77      0.97      0.83       240
weighted avg       0.97      0.95      0.95       240

F1 score: 69.77%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: Extra Trees
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for Extra Trees: {'max_depth': None, 'min_samples_split': 2, 'n_estimators': 200}
No description has been provided for this image
Model: ExtraTreesClassifier(n_estimators=200, n_jobs=-1, random_state=42)
              precision    recall  f1-score   support

           0       0.96      0.99      0.97       225
           1       0.62      0.33      0.43        15

    accuracy                           0.95       240
   macro avg       0.79      0.66      0.70       240
weighted avg       0.94      0.95      0.94       240

F1 score: 43.48%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: Extra Trees
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for Extra Trees: {'max_depth': None, 'min_samples_split': 2, 'n_estimators': 200}
No description has been provided for this image
Model: ExtraTreesClassifier(n_estimators=200, n_jobs=-1, random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.99      0.98       225
           1       0.70      0.47      0.56        15

    accuracy                           0.95       240
   macro avg       0.83      0.73      0.77       240
weighted avg       0.95      0.95      0.95       240

F1 score: 56.00%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: Extra Trees
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for Extra Trees: {'max_depth': None, 'min_samples_split': 2, 'n_estimators': 200}
No description has been provided for this image
Model: ExtraTreesClassifier(n_estimators=200, n_jobs=-1, random_state=42)
              precision    recall  f1-score   support

       False       0.96      1.00      0.98       225
        True       0.86      0.40      0.55        15

    accuracy                           0.96       240
   macro avg       0.91      0.70      0.76       240
weighted avg       0.95      0.96      0.95       240

F1 score: 54.55%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: Extra Trees
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for Extra Trees: {'max_depth': None, 'min_samples_split': 2, 'n_estimators': 200}
No description has been provided for this image
Model: ExtraTreesClassifier(n_estimators=200, n_jobs=-1, random_state=42)
              precision    recall  f1-score   support

       False       0.96      1.00      0.98       225
        True       0.83      0.33      0.48        15

    accuracy                           0.95       240
   macro avg       0.90      0.66      0.73       240
weighted avg       0.95      0.95      0.94       240

F1 score: 47.62%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: Gradient Boosting
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Gradient Boosting: {'learning_rate': 0.05, 'max_depth': 2, 'n_estimators': 300}
No description has been provided for this image
Model: GradientBoostingClassifier(learning_rate=0.05, max_depth=2, n_estimators=300,
                           random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.97      0.97       225
           1       0.53      0.53      0.53        15

    accuracy                           0.94       240
   macro avg       0.75      0.75      0.75       240
weighted avg       0.94      0.94      0.94       240

F1 score: 53.33%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: Gradient Boosting
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Gradient Boosting: {'learning_rate': 0.05, 'max_depth': 2, 'n_estimators': 300}
No description has been provided for this image
Model: GradientBoostingClassifier(learning_rate=0.05, max_depth=2, n_estimators=300,
                           random_state=42)
              precision    recall  f1-score   support

           0       0.98      0.96      0.97       225
           1       0.56      0.67      0.61        15

    accuracy                           0.95       240
   macro avg       0.77      0.82      0.79       240
weighted avg       0.95      0.95      0.95       240

F1 score: 60.61%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: Gradient Boosting
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Gradient Boosting: {'learning_rate': 0.1, 'max_depth': 3, 'n_estimators': 300}
No description has been provided for this image
Model: GradientBoostingClassifier(n_estimators=300, random_state=42)
              precision    recall  f1-score   support

       False       0.97      0.98      0.98       225
        True       0.67      0.53      0.59        15

    accuracy                           0.95       240
   macro avg       0.82      0.76      0.78       240
weighted avg       0.95      0.95      0.95       240

F1 score: 59.26%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: Gradient Boosting
-------------------------------------------------------------------------------
Fitting 5 folds for each of 8 candidates, totalling 40 fits
Better parameters for Gradient Boosting: {'learning_rate': 0.05, 'max_depth': 2, 'n_estimators': 300}
No description has been provided for this image
Model: GradientBoostingClassifier(learning_rate=0.05, max_depth=2, n_estimators=300,
                           random_state=42)
              precision    recall  f1-score   support

       False       0.97      0.96      0.97       225
        True       0.53      0.60      0.56        15

    accuracy                           0.94       240
   macro avg       0.75      0.78      0.77       240
weighted avg       0.95      0.94      0.94       240

F1 score: 56.25%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: HistGB
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for HistGB: {'learning_rate': 0.1, 'max_depth': None, 'max_leaf_nodes': 31}
No description has been provided for this image
Model: HistGradientBoostingClassifier(random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.97      0.97       225
           1       0.60      0.60      0.60        15

    accuracy                           0.95       240
   macro avg       0.79      0.79      0.79       240
weighted avg       0.95      0.95      0.95       240

F1 score: 60.00%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: HistGB
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for HistGB: {'learning_rate': 0.05, 'max_depth': None, 'max_leaf_nodes': 31}
No description has been provided for this image
Model: HistGradientBoostingClassifier(learning_rate=0.05, random_state=42)
              precision    recall  f1-score   support

           0       0.99      0.96      0.98       225
           1       0.60      0.80      0.69        15

    accuracy                           0.95       240
   macro avg       0.79      0.88      0.83       240
weighted avg       0.96      0.95      0.96       240

F1 score: 68.57%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: HistGB
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for HistGB: {'learning_rate': 0.1, 'max_depth': None, 'max_leaf_nodes': 31}
No description has been provided for this image
Model: HistGradientBoostingClassifier(random_state=42)
              precision    recall  f1-score   support

       False       0.98      0.97      0.97       225
        True       0.59      0.67      0.62        15

    accuracy                           0.95       240
   macro avg       0.78      0.82      0.80       240
weighted avg       0.95      0.95      0.95       240

F1 score: 62.50%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: HistGB
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
Better parameters for HistGB: {'learning_rate': 0.05, 'max_depth': None, 'max_leaf_nodes': 31}
No description has been provided for this image
Model: HistGradientBoostingClassifier(learning_rate=0.05, random_state=42)
              precision    recall  f1-score   support

       False       0.98      0.97      0.98       225
        True       0.61      0.73      0.67        15

    accuracy                           0.95       240
   macro avg       0.80      0.85      0.82       240
weighted avg       0.96      0.95      0.96       240

F1 score: 66.67%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: KNN
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.2s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.6s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.4s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.3s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
Better parameters for KNN: {'n_neighbors': 5, 'p': 2, 'weights': 'uniform'}
No description has been provided for this image
Model: KNeighborsClassifier()
              precision    recall  f1-score   support

           0       0.96      0.96      0.96       225
           1       0.43      0.40      0.41        15

    accuracy                           0.93       240
   macro avg       0.69      0.68      0.69       240
weighted avg       0.93      0.93      0.93       240

F1 score: 41.38%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: KNN
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.2s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END ...................................C=10, penalty=l2; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
Better parameters for KNN: {'n_neighbors': 5, 'p': 1, 'weights': 'distance'}
No description has been provided for this image
Model: KNeighborsClassifier(p=1, weights='distance')
              precision    recall  f1-score   support

           0       0.99      0.92      0.95       225
           1       0.41      0.80      0.55        15

    accuracy                           0.92       240
   macro avg       0.70      0.86      0.75       240
weighted avg       0.95      0.92      0.93       240

F1 score: 54.55%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: KNN
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_sam
ples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END ...................................C=10, penalty=l2; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.6s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.2s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.3s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.7s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.3s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.7s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
Better parameters for KNN: {'n_neighbors': 5, 'p': 2, 'weights': 'uniform'}
No description has been provided for this image
Model: KNeighborsClassifier()
              precision    recall  f1-score   support

       False       0.97      0.96      0.97       225
        True       0.50      0.53      0.52        15

    accuracy                           0.94       240
   macro avg       0.73      0.75      0.74       240
weighted avg       0.94      0.94      0.94       240

F1 score: 51.61%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: KNN
-------------------------------------------------------------------------------
Fitting 5 folds for each of 12 candidates, totalling 60 fits
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.2s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.5s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.3s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.9s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.2s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ........................C=1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l2; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.1s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=None, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.8s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ...learning_rate=0.1, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.1s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.1s
[CV] END .learning_rate=0.05, max_depth=6, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.1, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=2, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=20, min_samples_split=5, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.1s
[CV] END max_depth=None, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=10, min_samples_split=5, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=20, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=None, min_samples_split=2, n_estimators=300; total time=   0.6s
[CV] END max_depth=None, min_samples_split=5, n_estimators=100; total time=   0.2s
[CV] END max_depth=10, min_samples_split=2, n_estimators=100; total time=   0.1s
[CV] END max_depth=10, min_samples_split=5, n_estimators=100; total time=   0.3s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=300; total time=   0.4s
[CV] END max_depth=None, min_samples_split=10, n_estimators=300; total time=   0.3s
[CV] END max_depth=10, min_samples_split=2, n_estimators=300; total time=   0.5s
[CV] END max_depth=10, min_samples_split=10, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=2, n_estimators=100; total time=   0.2s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=20, min_samples_split=5, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .......................C=10, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .......................C=1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END .....................C=0.1, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=10, gamma=scale, kernel=rbf; total time=   0.0s
[CV] END ......................C=0.1, gamma=auto, kernel=rbf; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.0s
[CV] END .................................C=0.01, penalty=l2; total time=   0.0s
[CV] END ....................................C=1, penalty=l2; total time=   0.0s
[CV] END .................................C=0.01, penalty=l1; total time=   0.0s
[CV] END ..................................C=0.1, penalty=l2; total time=   0.0s
[CV] END ...................................C=10, penalty=l1; total time=   0.1s
[CV] END ..................................C=0.1, penalty=l1; total time=   0.0s
[CV] END ....................................C=1, penalty=l1; total time=   0.1s
[CV] END ...................................C=10, penalty=l2; total time=   0.2s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.6s
[CV] END max_depth=10, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=200; total time=   0.3s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=20, min_samples_split=10, n_estimators=400; total time=   0.5s
[CV] END max_depth=None, min_samples_split=10, n_estimators=400; total time=   0.7s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=400; total time=   0.6s
[CV] END max_depth=None, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=10, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=2, n_estimators=200; total time=   0.4s
[CV] END max_depth=20, min_samples_split=10, n_estimators=200; total time=   0.4s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.5s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.1s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=300; total time=   0.5s
[CV] END ...learning_rate=0.1, max_depth=3, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=2, n_estimators=100; total time=   0.2s
[CV] END ..learning_rate=0.05, max_depth=3, n_estimators=300; total time=   0.6s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=63; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=63; total time=   0.1s
[CV] END learning_rate=0.05, max_depth=None, max_leaf_nodes=31; total time=   0.2s
[CV] END learning_rate=0.05, max_depth=12, max_leaf_nodes=63; total time=   0.2s
[CV] END ..learning_rate=0.1, max_depth=6, max_leaf_nodes=31; total time=   0.2s
[CV] END .learning_rate=0.1, max_depth=12, max_leaf_nodes=31; total time=   0.2s
[CV] END ................n_neighbors=5, p=2, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=5, p=1, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=2, weights=distance; total time=   0.0s
[CV] END ...............n_neighbors=30, p=2, weights=uniform; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ..............n_neighbors=15, p=1, weights=distance; total time=   0.0s
[CV] END ..............n_neighbors=30, p=1, weights=distance; total time=   0.0s
[CV] END ................n_neighbors=5, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=15, p=1, weights=uniform; total time=   0.0s
[CV] END ...............n_neighbors=30, p=1, weights=uniform; total time=   0.0s
Better parameters for KNN: {'n_neighbors': 5, 'p': 2, 'weights': 'distance'}
No description has been provided for this image
Model: KNeighborsClassifier(weights='distance')
              precision    recall  f1-score   support

       False       0.99      0.91      0.95       225
        True       0.39      0.87      0.54        15

    accuracy                           0.91       240
   macro avg       0.69      0.89      0.75       240
weighted avg       0.95      0.91      0.92       240

F1 score: 54.17%
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
-------------------------- Original dataset -----------------------------------
Model: Linear SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Better parameters for Linear SVM: {'C': 10}
No description has been provided for this image
Model: LinearSVC(C=10, random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.97      0.97       225
           1       0.60      0.60      0.60        15

    accuracy                           0.95       240
   macro avg       0.79      0.79      0.79       240
weighted avg       0.95      0.95      0.95       240

F1 score: 60.00%
-------------------------------------------------------------------------------
-------------------------- Random oversampled ---------------------------------
Model: Linear SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Better parameters for Linear SVM: {'C': 0.1}
No description has been provided for this image
Model: LinearSVC(C=0.1, random_state=42)
              precision    recall  f1-score   support

           0       1.00      0.93      0.97       225
           1       0.50      1.00      0.67        15

    accuracy                           0.94       240
   macro avg       0.75      0.97      0.82       240
weighted avg       0.97      0.94      0.95       240

F1 score: 66.67%
-------------------------------------------------------------------------------
-------------------------- Original dataset one hot encoded -------------------
Model: Linear SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Better parameters for Linear SVM: {'C': 1}
No description has been provided for this image
Model: LinearSVC(C=1, random_state=42)
              precision    recall  f1-score   support

       False       0.97      0.97      0.97       225
        True       0.57      0.53      0.55        15

    accuracy                           0.95       240
   macro avg       0.77      0.75      0.76       240
weighted avg       0.94      0.95      0.94       240

F1 score: 55.17%
-------------------------------------------------------------------------------
-------------------------- Random oversampled one hot encoded -----------------
Model: Linear SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Better parameters for Linear SVM: {'C': 0.1}
No description has been provided for this image
Model: LinearSVC(C=0.1, random_state=42)
              precision    recall  f1-score   support

       False       1.00      0.93      0.96       225
        True       0.48      1.00      0.65        15

    accuracy                           0.93       240
   macro avg       0.74      0.96      0.81       240
weighted avg       0.97      0.93      0.94       240

F1 score: 65.22%
-------------------------------------------------------------------------------

Model choice¶

In [18]:
model = "SVM"

print('-------------------------- Original dataset -----------------------------------')
print(f'Model: {model}')
print('-------------------------------------------------------------------------------')
X_train, X_test, y_train, y_test = preprocess(aux, drops, numerical_features, "Fail")
tmodel = train_model(X_train, y_train, model)
predict_and_evaluate(tmodel, X_test, y_test)
-------------------------- Original dataset -----------------------------------
Model: SVM
-------------------------------------------------------------------------------
Fitting 5 folds for each of 6 candidates, totalling 30 fits
Better parameters for SVM: {'C': 10, 'gamma': 'scale', 'kernel': 'rbf'}
No description has been provided for this image
Model: SVC(C=10, random_state=42)
              precision    recall  f1-score   support

           0       0.99      0.97      0.98       225
           1       0.63      0.80      0.71        15

    accuracy                           0.96       240
   macro avg       0.81      0.88      0.84       240
weighted avg       0.96      0.96      0.96       240

F1 score: 70.59%

Task 5¶

  • Analyze variable importance.

  • After developing a model, how can we determine which variables had the greatest impact on the prediction?

In [19]:
X_train, X_test, y_train, y_test = preprocess(aux, drops, numerical_features, "Fail")

tmodel = RandomForestClassifier(max_depth=10, min_samples_split=10, 
                                n_estimators=300, random_state=42)

tmodel.fit(X_train, y_train)

predict_and_evaluate(tmodel, X_test, y_test)

def plot_feature_importance(model, X_train):

    importances = model.feature_importances_

    feature_names = X_train.columns

    feature_importance_df = pd.DataFrame({
        'Feature': feature_names,
        'Importance': importances
    }).sort_values(by='Importance', ascending=False)

    plt.figure(figsize=(10, 6))
    plt.barh(feature_importance_df['Feature'], feature_importance_df['Importance'], color='royalblue')
    plt.xlabel('Importance')
    plt.ylabel('Feature')
    plt.title('Feature importance - Random Forest')
    plt.gca().invert_yaxis()
    plt.show()

plot_feature_importance(tmodel, X_train)
No description has been provided for this image
Model: RandomForestClassifier(max_depth=10, min_samples_split=10, n_estimators=300,
                       random_state=42)
              precision    recall  f1-score   support

           0       0.97      0.98      0.97       225
           1       0.64      0.47      0.54        15

    accuracy                           0.95       240
   macro avg       0.80      0.72      0.76       240
weighted avg       0.94      0.95      0.95       240

F1 score: 53.85%
No description has been provided for this image